Product Code Database
Example Keywords: house -sweater $68-195
barcode-scavenger
   » » Wiki: Total Variation
Tag Wiki 'Total Variation'.
Tag

In , the total variation identifies several slightly different concepts, related to the ( or global) structure of the of a function or a measure. For a continuous function f, defined on an interval a, ⊂ R, its total variation on the interval of definition is a measure of the one-dimensional of the curve with parametric equation xf( x), for xa,. Functions whose total variation is finite are called functions of bounded variation.


Historical note
The concept of total variation for functions of one real variable was first introduced by in the paper .According to . He used the new concept in order to prove a convergence theorem for of discontinuous periodic functions whose variation is bounded. The extension of the concept to functions of more than one variable however is not simple for various reasons.


Definitions

Total variation for functions of one real variable
The '''total variation''' of a [[real|real number]]-valued (or more generally [[complex|complex number]]-valued) function f, defined on an interval  [a , b] \subset \mathbb{R} is the quantity
     

V_a^b(f)=\sup_{\mathcal{P}} \sum_{i=0}^{n_P-1} | f(x_{i+1})-f(x_i) |,

where the runs over the set of all partitions \mathcal{P} = \left\{P=\{ x_0, \dots , x_{n_P}\} \mid P\text{ is a partition of } a,b \right\} of the given interval. Which means that a = x_{0} < x_{1} < ... < x_{n_{P}} = b.


Total variation for functions of n > 1 real variables
(2026). 9780198502456, Oxford University Press. .
Let Ω be an of R n. Given a function f belonging to L1( Ω), the total variation of f in Ω is defined as

V(f,\Omega):=\sup\left\{\int_\Omega f(x) \operatorname{div} \phi(x) \, \mathrm{d}x \colon \phi\in C_c^1(\Omega,\mathbb{R}^n),\ \Vert \phi\Vert_{L^\infty(\Omega)}\le 1\right\},

where

  • C_c^1(\Omega,\mathbb{R}^n) is the set of vector functions of compact support contained in \Omega,
  • \Vert\;\Vert_{L^\infty(\Omega)} is the essential supremum norm, and
  • \operatorname{div} is the operator.
This definition does not require that the domain \Omega \subseteq \mathbb{R}^n of the given function be a .


Total variation in measure theory

Classical total variation definition
Following , consider a \mu on a (X,\Sigma): then it is possible to define two \overline{\mathrm{W}}(\mu,\cdot) and \underline{\mathrm{W}}(\mu,\cdot), respectively called upper variation and lower variation, as follows

\overline{\mathrm{W}}(\mu,E)=\sup\left\{\mu(A)\mid A\in\Sigma\text{ and }A\subset E \right\}\qquad\forall E\in\Sigma
\underline{\mathrm{W}}(\mu,E)=\inf\left\{\mu(A)\mid A\in\Sigma\text{ and }A\subset E \right\}\qquad\forall E\in\Sigma

clearly

\overline{\mathrm{W}}(\mu,E)\geq 0 \geq \underline{\mathrm{W}}(\mu,E)\qquad\forall E\in\Sigma

The '''variation''' (also called '''absolute variation''') of the signed measure \mu is the set function
     

|\mu|(E)=\overline{\mathrm{W}}(\mu,E)+\left|\underline{\mathrm{W}}(\mu,E)\right|\qquad\forall E\in\Sigma

and its total variation is defined as the value of this measure on the whole space of definition, i.e.

\|\mu\|=|\mu|(X)


Modern definition of total variation norm
uses upper and lower variations to prove the Hahn–Jordan decomposition: according to his version of this theorem, the upper and lower variation are respectively a [[non-negative]] and a [[non-positive]] measure. Using a more modern notation, define
     

\mu^+(\cdot)=\overline{\mathrm{W}}(\mu,\cdot)\,,
\mu^-(\cdot)=-\underline{\mathrm{W}}(\mu,\cdot)\,,

Then \mu^+ and \mu^- are two non-negative measures such that

\mu=\mu^+-\mu^-
|\mu|=\mu^++\mu^-

The last measure is sometimes called, by abuse of notation, total variation measure.


Total variation norm of complex measures
If the measure \mu is i.e. is a , its upper and lower variation cannot be defined and the Hahn–Jordan decomposition theorem can only be applied to its real and imaginary parts. However, it is possible to follow and define the total variation of the complex-valued measure \mu as follows

The '''variation''' of the complex-valued measure \mu is the [[set function]]
     

|\mu|(E)=\sup_\pi \sum_{A\isin\pi} |\mu(A)|\qquad\forall E\in\Sigma

where the is taken over all partitions \pi of a E into a countable number of disjoint measurable subsets.

This definition coincides with the above definition |\mu|=\mu^++\mu^- for the case of real-valued signed measures.


Total variation norm of vector-valued measures
The variation so defined is a (see ) and coincides with the one defined by when \mu is a : its total variation is defined as above. This definition works also if \mu is a : the variation is then defined by the following formula

|\mu|(E) = \sup_\pi \sum_{A\isin\pi} \|\mu(A)\|\qquad\forall E\in\Sigma

where the supremum is as above. This definition is slightly more general than the one given by since it requires only to consider finite partitions of the space X: this implies that it can be used also to define the total variation on .


Total variation of probability measures
The total variation of any probability measure is exactly one, therefore it is not interesting as a means of investigating the properties of such measures. However, when μ and ν are probability measures, the total variation distance of probability measures can be defined as \| \mu - \nu \| where the norm is the total variation norm of signed measures. Using the property that (\mu-\nu)(X)=0, we eventually arrive at the equivalent definition

\|\mu-\nu\| = |\mu-\nu|(X)=2 \sup\left\{\,\left|\mu(A)-\nu(A)\right| : A\in \Sigma\,\right\}

and its values are non-trivial. The factor 2 above is usually dropped (as is the convention in the article total variation distance of probability measures). Informally, this is the largest possible difference between the probabilities that the two probability distributions can assign to the same event. For a categorical distribution it is possible to write the total variation distance as follows

\delta(\mu,\nu) = \sum_x \left| \mu(x) - \nu(x) \right|\;.

It may also be normalized to values in 0, by halving the previous definition as follows

\delta(\mu,\nu) = \frac{1}{2}\sum_x \left| \mu(x) - \nu(x) \right|


Basic properties

Total variation of differentiable functions
The total variation of a C^1(\overline{\Omega}) function f can be expressed as an involving the given function instead of as the of the functionals of definitions and .


The form of the total variation of a differentiable function of one variable
The '''total variation''' of a differentiable function f, defined on an interval  [a , b] \subset \mathbb{R}, has the following expression if f' is Riemann integrable
     

V_a^b(f) = \int _a^b |f'(x)|\mathrm{d}x

If f is differentiable and monotonic, then the above simplifies to

V_a^b(f) = |f(a) - f(b)|

For any differentiable function f, we can decompose the domain interval a,b, into subintervals a,a_1, a_1,a_2, \dots, a_N,b (with a) in which f is locally monotonic, then the total variation of f over a,b can be written as the sum of local variations on those subintervals:

\begin{align} V_a^b(f) &= V_a^{a_1}(f) + V_{a_1}^{a_2}(f) + \, \cdots \, +V_{a_N}^b(f)\\0.3em &=|f(a)-f(a_1)|+|f(a_1)-f(a_2)|+ \,\cdots \, + |f(a_N)-f(b)| \end{align}


The form of the total variation of a differentiable function of several variables
Given a C^1(\overline{\Omega}) function f defined on a [[bounded|bounded set]] [[open set]] \Omega \subseteq \mathbb{R}^n, with \partial \Omega  of class C^1,  the '''total variation of f''' has the following expression
     

V(f,\Omega) = \int_\Omega \left|\nabla f(x) \right| \mathrm{d}x .


Proof
The first step in the proof is to first prove an equality which follows from the Gauss–Ostrogradsky theorem.


Lemma
Under the conditions of the theorem, the following equality holds:
\int_\Omega f\operatorname{div}\varphi = -\int_\Omega\nabla f\cdot\varphi


Proof of the lemma
From the Gauss–Ostrogradsky theorem:
\int_\Omega \operatorname{div}\mathbf R = \int_{\partial\Omega}\mathbf R\cdot \mathbf n
by substituting \mathbf R:= f\mathbf\varphi, we have:

\int_\Omega\operatorname{div}\left(f\mathbf\varphi\right) =
\int_{\partial\Omega}\left(f\mathbf\varphi\right)\cdot\mathbf n where \mathbf\varphi is zero on the border of \Omega by definition:
\int_\Omega\operatorname{div}\left(f\mathbf\varphi\right)=0
\int_\Omega \partial_{x_i} \left(f\mathbf\varphi_i\right)=0
\int_\Omega \mathbf\varphi_i\partial_{x_i} f + f\partial_{x_i}\mathbf\varphi_i=0
\int_\Omega f\partial_{x_i}\mathbf\varphi_i = - \int_\Omega \mathbf\varphi_i\partial_{x_i} f
\int_\Omega f\operatorname{div} \mathbf\varphi = - \int_\Omega \mathbf\varphi\cdot\nabla f


Proof of the equality
Under the conditions of the theorem, from the lemma we have:
\int_\Omega f\operatorname{div} \mathbf\varphi
= - \int_\Omega \mathbf\varphi\cdot\nabla f \leq \left| \int_\Omega \mathbf\varphi\cdot\nabla f \right| \leq \int_\Omega \left|\mathbf\varphi\right|\cdot\left|\nabla f\right| \leq \int_\Omega \left|\nabla f\right| in the last part \mathbf\varphi could be omitted, because by definition its essential supremum is at most one.

On the other hand, we consider \theta_N:=-\mathbb I_{\left-N,N\right}\mathbb I_{\{\nabla f\ne 0\}}\frac{\nabla f}{\left|\nabla f\right|} and \theta^*_N which is the up to \varepsilon approximation of \theta_N in C^1_c with the same integral. We can do this since C^1_c is dense in L^1 . Now again substituting into the lemma:

\begin{align}
&\lim_{N\to\infty}\int_\Omega f\operatorname{div}\theta^*_N \\4pt &= \lim_{N\to\infty}\int_{\{\nabla f\ne 0\}}\mathbb I_{\left-N,N\right}\nabla f\cdot\frac{\nabla f}{\left|\nabla f\right|} \\4pt &= \lim_{N\to\infty}\int_{\left-N,N\right\cap{\{\nabla f\ne 0\}}} \nabla f\cdot\frac{\nabla f}{\left|\nabla f\right|} \\4pt &= \int_\Omega\left|\nabla f\right| \end{align} This means we have a convergent sequence of \int_\Omega f \operatorname{div} \mathbf\varphi that tends to \int_\Omega\left|\nabla f\right| as well as we know that \int_\Omega f\operatorname{div}\mathbf\varphi \leq \int_\Omega\left|\nabla f\right| . Q.E.D.

It can be seen from the proof that the supremum is attained when

\varphi\to \frac{-\nabla f}{\left|\nabla f\right|}.

The function f is said to be of bounded variation precisely if its total variation is finite.


Total variation of a measure
The total variation is a norm defined on the space of measures of bounded variation. The space of measures on a σ-algebra of sets is a , called the , relative to this norm. It is contained in the larger Banach space, called the , consisting of finitely additive (as opposed to countably additive) measures, also with the same norm. The distance function associated to the norm gives rise to the total variation distance between two measures μ and ν.

For finite measures on R, the link between the total variation of a measure μ and the total variation of a function, as described above, goes as follows. Given μ, define a function \varphi\colon \mathbb{R}\to \mathbb{R} by

\varphi(t) = \mu((-\infty,t])~.
Then, the total variation of the signed measure μ is equal to the total variation, in the above sense, of the function \varphi. In general, the total variation of a signed measure can be defined using Jordan's decomposition theorem by
\|\mu\|_{TV} = \mu_+(X) + \mu_-(X)~,
for any signed measure μ on a measurable space (X,\Sigma).


Applications
Total variation can be seen as a -valued functional defined on the space of functions (for the case of functions of one variable) or on the space of integrable functions (for the case of functions of several variables). As a functional, total variation finds applications in several branches of mathematics and engineering, like , numerical analysis, and calculus of variations, where the solution to a certain problem has to minimize its value. As an example, use of the total variation functional is common in the following two kind of problems

  • Numerical analysis of differential equations: it is the science of finding approximate solutions to differential equations. Applications of total variation to these problems are detailed in the article " total variation diminishing"
  • Image denoising Https://arxiv.org/pdf/1603.09599 Retrieved 12/15/2024 in , denoising is a collection of methods used to reduce the in an reconstructed from data obtained by electronic means, for example data transmission or . " Total variation denoising" is the name for the application of total variation to image noise reduction; further details can be found in the papers of and . A sensible extension of this model to colour images, called Colour TV, can be found in .


See also
  • Bounded variation
  • Total variation diminishing
  • Total variation denoising
  • Quadratic variation
  • Total variation distance of probability measures
  • Kolmogorov–Smirnov test
  • Anisotropic diffusion


Notes

Historical references
  • .
  • .
  • .
  • .
  • .
  • .
  • .
  • (available at ). This is, according to Boris Golubov, the first paper on functions of bounded variation.
  • .
  • . The paper containing the first proof of Vitali covering theorem.


External links
One variable
  • "

One and more variables

Measure theory


Applications
  • (a work dealing with total variation application in denoising problems for ).

  • .

  • .

Page 1 of 1
1
Page 1 of 1
1

Account

Social:
Pages:  ..   .. 
Items:  .. 

Navigation

General: Atom Feed Atom Feed  .. 
Help:  ..   .. 
Category:  ..   .. 
Media:  ..   .. 
Posts:  ..   ..   .. 

Statistics

Page:  .. 
Summary:  .. 
1 Tags
10/10 Page Rank
5 Page Refs